PicSOM Experiments in TRECVID 2009 Workshop draft

نویسندگان

  • Mats Sjöberg
  • Ville Viitaniemi
  • Markus Koskela
  • Jorma Laaksonen
چکیده

Our experiments in TRECVID 2009 include participation in the high-level feature extraction and automatic search tasks. In the high-level feature extraction task, we used a feature fusion-based general system architecture utilizing a large number of SVM detectors, followed by a post-processing stage utilizing the concepts’ temporal and inter-concept co-occurrences. We submitted the following six runs: • PicSOM.base: Baseline run using our SOM-based HLF detection method • PicSOM.A-ngram: Baseline SVM-based run using HLF-wise geometric mean fusion and temporal n-gram post-processing • PicSOM.B-ngram: As previous, but includes also early fusion, multi-fold SFBS fusion, and more elaborate SVM training • PicSOM.E-ngram: As previous, but includes two-stage fusion utilizing cross-concept co-occurrence • PicSOM.spec-ngram: A run where the used method was selected for each HLF separately using cross-validation • PicSOM.spec-any: As previous, but the post-processing used also clustering-based inter-concept co-occurrence analysis The results show that feature fusion can consistently outperform all single features, multi-fold SFBS performed best of the tested fusion methods, and that temporal n-gram analysis is beneficial. Early fusion, and post-processing based on inter-concept co-occurrences did not improve the performance. In the search task, we concentrated on the fully-automatic runs and standard search task. We combined ASR/MT text search and concept-based retrieval. If none of the concept models could be matched with the query, we used content-based retrieval based on the video and image examples instead. We submitted the following ten fully-automatic runs: • F_A_N_PicSOM_1_10: text search baseline • F_A_N_PicSOM_2_9: visual baseline • F_A_N_PicSOM_3_8: own concepts • F_A_N_PicSOM_4_7: own concepts + text search • F_A_N_PicSOM_5_6: donated concepts • F_A_N_PicSOM_6_5: donated concepts + text search • F_A_N_PicSOM_7_4: own + donated concepts • F_A_N_PicSOM_8_3: own + donated concepts + text search • F_A_N_PicSOM_9_2: own + donated (dupl.) concepts • F_A_N_PicSOM_10_1: own + donated (dupl.) concepts + text search In the above list, “own” concepts refer to our own HLF detectors and “donated” concepts consist of MediaMill (MM) concepts + CU-VIREO374 concepts. In other than the last two runs, CU-VIREO374 are only used for words for which no MediaMill concept could be matched. The results show again that concept-based retrieval performed better than content-based search alone. Text search made a small improvement in combination with other modalities, but performed really badly on its own. Concept-selection was done both with word-matching and example-based matching, i.e. selecting concepts based on how well they would fit our own concept models.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Concept-based Video Search with the Picsom Multi- Media Retrieval System

In this report we describe the structure of the PicSOM multimedia retrieval system and elaborate on its automatic concept detection and video search subsystems. We evaluate several alternative techniques for implementing these two components of the PicSOM system in a comprehensive series of experiments employing the large-scale setups of the TRECVID video retrieval evaluation campaigns of 2008 ...

متن کامل

PicSOM Experiments in TRECVID 2014

Our experiments in TRECVID 2014 include successful participation in the Semantic Indexing (SIN) task and unsuccessful participation in the Multimedia Event Detection (MED) and Multimedia Event Recounting (MER) tasks. In semantic indexing, we participated in the main task only. We extended our last year’s set of features with SIFT descriptors encoded with Fisher vectors and VLAD, and a total of ...

متن کامل

PicSOM Experiments in TRECVID 2006

Our experiments in TRECVID 2006 include participation in the shot boundary detection, high-level feature extraction, and search tasks, using a common system framework based on multiple parallel Self-Organizing Maps (SOMs). In the shot boundary detection task we projected feature vectors calculated from successive frames on parallel SOMs and monitored the trajectories to detect the shot boundari...

متن کامل

PicSOM Experiments in TRECVID 2010 Workshop draft

Our experiments in TRECVID 2010 include participation in the semantic indexing and known-item search tasks. In the semantic indexing task we implemented SVM-based classifiers on five different low-level visual features extracted from the keyframes. In addition to the main keyframes provided by NIST, we also extracted and analysed additional frames from longer shots. The feature-wise classifiers...

متن کامل

Video Searching and Browsing Using ViewFinder: Interactive Search Experiments for TRECVID-2003

This research project explores the topic of video information retrieval in conjunction with the task definitions and data provided by the Text REtrieval Conference’s (TREC) 2003 Video Workshop (TRECVID-2003). Included in this paper, we discuss our processes and various phases in participating with TRECVID-2003. Specific sections discussed include database development, data indexing and retrieva...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009